In automated industrial scenarios, the amount of time series log data generated by a large number of industrial devices has exploded, and the demand for access to time series data in business scenarios has further increased. Although HBase, a distributed column family database, can store industrial time series big data, the existing strategies cannot meet the specific access requirements of industrial time series data well because the correlation between data and access behavior characteristics in specific business scenarios is not considered. In view of the above problem, based on the distributed storage system HBase, and using the correlation between data and access behavior characteristics in industrial scenarios, a distributed storage performance optimization strategy for massive industrial time series data was proposed. Aiming at the load tilt problem caused by characteristics of industrial time series data, a load balancing optimization strategy based on hot and cold data partition and access behavior classification was proposed. The data were classified into cold and hot ones by using a Logistic Regression (LR) model, and the hot data were distributed and stored in different nodes. In addition, in order to further reduce the cross-node communication overhead in storage cluster and improve the query efficiency of the high-dimensional index of industrial time series data, a strategy of putting the index and main data into a same Region was proposed. By designing the index RowKey field and splicing rules, the index was stored with its corresponding main data in the same Region. Experimental results on real industrial time series data show that the data load distribution tilt degree is reduced by 28.5% and the query efficiency is improved by 27.7% after introducing the optimization strategy, demonstrating the proposed strategy can mine access patterns for specific time series data effectively, distribute load reasonably, reduce data access overhead, and meet access requirements for specific time series big data.
Point cloud data has sparsity, irregularity, and permutation invariance, and lacks topological information, which makes it difficult to extract features of point cloud. Therefore, a Siamese Adaptive Graph Convolution Algorithm (SAGCA) was proposed for point cloud classification and segmentation. Firstly, the topological relationships between irregular and sparse point cloud features were mined by constructing feature relationship graph. Then, the Siamese composition idea of sharing convolution learning weights was introduced to ensure the permutation invariance of point cloud data and make the topological relationship expression more accurate. Finally, SAGCA was combined with various deep learning networks for processing point cloud data by both global and local combination methods, thereby enhancing the feature extraction ability of the network. Comparison results with PointNet++ benchmark network of the classification, object part segmentation and scene semantic segmentation experiments on ScanObjectNN, ShapeNetPart and S3DIS datasets, respectively, show that, based on the same dataset and evaluation criteria, SAGCA has the class mean Accuracy (mAcc) of classification increased by 2.80 percentage points, the overall class average Intersection over Union (IoU) of part segmentation increased by 2.31 percentage points, and the class mean Intersection over Union (mIoU) of scene semantic segmentation increased by 2.40 percentage points, verifying that SAGCA can effectively enhance the feature extraction ability of the network and is suitable for multiple point cloud classification and segmentation tasks.